radar sensor
- Europe > Switzerland (0.04)
- Europe > Netherlands > South Holland > Delft (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Transportation > Ground > Road (1.00)
- Information Technology (1.00)
- Automobiles & Trucks (1.00)
- (2 more...)
Radar-Based Odometry for Low-Speed Driving
Diener, Luis, Kalkkuhl, Jens, Enzweiler, Markus
Abstract--We address automotive odometry for low-speed driving and parking, where centimeter-level accuracy is required due to tight spaces and nearby obstacles. Traditional methods using inertial-measurement units and wheel encoders require vehicle-specific calibration, making them costly for consumer-grade vehicles. T o overcome this, we propose a radar-based simultaneous localization and mapping (SLAM) approach that fuses inertial and 4D radar measurements. Our approach tightly couples feature positions and Doppler velocities for accurate localization and robust data association. Key contributions include a tightly coupled radar-Doppler extended Kalman filter, multi-radar support and an information-based feature-pruning strategy. Experiments using both proprietary and public datasets demonstrate high-accuracy localization during low-speed driving. Accurate relative localization is critical for automated parking applications, where the vehicle executes low-speed maneuvers in complex environments. Unlike highway or urban driving, parking scenarios demand centimeter-level accuracy due to space constraints and the proximity to surrounding obstacles.
- Europe > Germany (0.14)
- Europe > Monaco (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- (4 more...)
- Transportation > Ground > Road (0.46)
- Automobiles & Trucks (0.46)
- Europe > Switzerland (0.04)
- Europe > Netherlands > South Holland > Delft (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Transportation > Ground > Road (1.00)
- Information Technology (1.00)
- Automobiles & Trucks (1.00)
- (2 more...)
Gesture-Based Robot Control Integrating Mm-wave Radar and Behavior Trees
Song, Yuqing, Tonola, Cesare, Savazzi, Stefano, Kianoush, Sanaz, Pedrocchi, Nicola, Sigg, Stephan
As robots become increasingly prevalent in both homes and industrial settings, the demand for intuitive and efficient human-machine interaction continues to rise. Gesture recognition offers an intuitive control method that does not require physical contact with devices and can be implemented using various sensing technologies. Wireless solutions are particularly flexible and minimally invasive. While camera-based vision systems are commonly used, they often raise privacy concerns and can struggle in complex or poorly lit environments. In contrast, radar sensing preserves privacy, is robust to occlusions and lighting, and provides rich spatial data such as distance, relative velocity, and angle. We present a gesture-controlled robotic arm using mm-wave radar for reliable, contactless motion recognition. Nine gestures are recognized and mapped to real-time commands with precision. Case studies are conducted to demonstrate the system practicality, performance and reliability for gesture-based robotic manipulation. Unlike prior work that treats gesture recognition and robotic control separately, our system unifies both into a real-time pipeline for seamless, contactless human-robot interaction.
Robust Multimodal Learning Framework For Intake Gesture Detection Using Contactless Radar and Wearable IMU Sensors
Wang, Chunzhuo, Hallez, Hans, Vanrumste, Bart
Automated food intake gesture detection plays a vital role in dietary monitoring, enabling objective and continuous tracking of eating behaviors to support better health outcomes. Wrist-worn inertial measurement units (IMUs) have been widely used for this task with promising results. More recently, contactless radar sensors have also shown potential. This study explores whether combining wearable and contactless sensing modalities through multimodal learning can further improve detection performance. We also address a major challenge in multimodal learning: reduced robustness when one modality is missing. To this end, we propose a robust multimodal temporal convolutional network with cross-modal attention (MM-TCN-CMA), designed to integrate IMU and radar data, enhance gesture detection, and maintain performance under missing modality conditions. A new dataset comprising 52 meal sessions (3,050 eating gestures and 797 drinking gestures) from 52 participants is developed and made publicly available. Experimental results show that the proposed framework improves the segmental F1-score by 4.3% and 5.2% over unimodal Radar and IMU models, respectively. Under missing modality scenarios, the framework still achieves gains of 1.3% and 2.4% for missing radar and missing IMU inputs. This is the first study to demonstrate a robust multimodal learning framework that effectively fuses IMU and radar data for food intake gesture detection.
- North America > United States (0.14)
- Europe > Belgium > Flanders > Flemish Brabant > Leuven (0.05)
- Information Technology (1.00)
- Health & Medicine > Consumer Health (1.00)
Privacy-aware IoT Fall Detection Services For Aging in Place
Lakhdari, Abdallah, Li, Jiajie, Abusafia, Amani, Bouguettaya, Athman
--Fall detection is critical to support the growing elderly population, projected to reach 2.1 billion by 2050. However, existing methods often face data scarcity challenges or compromise privacy. We propose a novel IoT -based Fall Detection as a Service (FDaaS) framework to assist the elderly in living independently and safely by accurately detecting falls. We address the challenges of data scarcity by utilizing a Fall Detection Generative Pre-trained Transformer (FD-GPT) that uses augmentation techniques. We developed a protocol to collect a comprehensive dataset of the elderly daily activities and fall events. This resulted in a real dataset that carefully mimics the elderly's routine. We rigorously evaluate and compare various models using this dataset. Experimental results show our approach achieves 90.72% accuracy and 89.33% precision in distinguishing between fall events and regular activities of daily living. The Internet of Things (IoT) enables everyday physical objects, or "things," to be connected to the Internet [1]. These objects are often equipped with pervasive intelligence capabilities. IoT devices' capabilities may be abstracted as IoT services [2]. An IoT service has a set of functional and non-functional, i.e., quality of service (QoS) properties.
- Oceania > Australia > Victoria > Melbourne (0.04)
- Oceania > Australia > New South Wales > Sydney (0.04)
- Europe > Switzerland (0.04)
- Health & Medicine > Consumer Health (0.47)
- Information Technology > Smart Houses & Appliances (0.36)
Learning Point Correspondences In Radar 3D Point Clouds For Radar-Inertial Odometry
Michalczyk, Jan, Weiss, Stephan, Steinbrener, Jan
Using 3D point clouds in odometry estimation in robotics often requires finding a set of correspondences between points in subsequent scans. While there are established methods for point clouds of sufficient quality, state-of-the-art still struggles when this quality drops. Thus, this paper presents a novel learning-based framework for predicting robust point correspondences between pairs of noisy, sparse and unstructured 3D point clouds from a light-weight, low-power, inexpensive, consumer-grade System-on-Chip (SoC) Frequency Modulated Continuous Wave (FMCW) radar sensor. Our network is based on the transformer architecture which allows leveraging the attention mechanism to discover pairs of points in consecutive scans with the greatest mutual affinity. The proposed network is trained in a self-supervised way using set-based multi-label classification cross-entropy loss, where the ground-truth set of matches is found by solving the Linear Sum Assignment (LSA) optimization problem, which avoids tedious hand annotation of the training data. Additionally, posing the loss calculation as multi-label classification permits supervising on point correspondences directly instead of on odometry error, which is not feasible for sparse and noisy data from the SoC radar we use. We evaluate our method with an open-source state-of-the-art Radar-Inertial Odometry (RIO) framework in real-world Unmanned Aerial Vehicle (UAV) flights and with the widely used public Coloradar dataset. Evaluation shows that the proposed method improves the position estimation accuracy by over 14 % and 19 % on average, respectively. The open source code and datasets can be found here: https://github.com/aau-cns/radar_transformer.
- Information Technology > Robotics & Automation (0.53)
- Aerospace & Defense > Aircraft (0.53)
Robust High-Speed State Estimation for Off-road Navigation using Radar Velocity Factors
Nissov, Morten, Edlund, Jeffrey A., Spieler, Patrick, Padgett, Curtis, Alexis, Kostas, Khattak, Shehryar
Enabling robot autonomy in complex environments for mission critical application requires robust state estimation. Particularly under conditions where the exteroceptive sensors, which the navigation depends on, can be degraded by environmental challenges thus, leading to mission failure. It is precisely in such challenges where the potential for FMCW radar sensors is highlighted: as a complementary exteroceptive sensing modality with direct velocity measuring capabilities. In this work we integrate radial speed measurements from a FMCW radar sensor, using a radial speed factor, to provide linear velocity updates into a sliding-window state estimator for fusion with LiDAR pose and IMU measurements. We demonstrate that this augmentation increases the robustness of the state estimator to challenging conditions present in the environment and the negative effects they can pose to vulnerable exteroceptive modalities. The proposed method is extensively evaluated using robotic field experiments conducted using an autonomous, full-scale, off-road vehicle operating at high-speeds (~12 m/s) in complex desert environments. Furthermore, the robustness of the approach is demonstrated for cases of both simulated and real-world degradation of the LiDAR odometry performance along with comparison against state-of-the-art methods for radar-inertial odometry on public datasets.
- North America > United States > New York (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Europe > Norway > Central Norway > Trøndelag > Trondheim (0.04)
Exploring Domain Shift on Radar-Based 3D Object Detection Amidst Diverse Environmental Conditions
Zhang, Miao, Abdulatif, Sherif, Loesch, Benedikt, Altmann, Marco, Schwarz, Marius, Yang, Bin
The rapid evolution of deep learning and its integration with autonomous driving systems have led to substantial advancements in 3D perception using multimodal sensors. Notably, radar sensors show greater robustness compared to cameras and lidar under adverse weather and varying illumination conditions. This study delves into the often-overlooked yet crucial issue of domain shift in 4D radar-based object detection, examining how varying environmental conditions, such as different weather patterns and road types, impact 3D object detection performance. Our findings highlight distinct domain shifts across various weather scenarios, revealing unique dataset sensitivities that underscore the critical role of radar point cloud generation. Additionally, we demonstrate that transitioning between different road types, especially from highways to urban settings, introduces notable domain shifts, emphasizing the necessity for diverse data collection across varied road environments. To the best of our knowledge, this is the first comprehensive analysis of domain shift effects on 4D radar-based object detection. We believe this empirical study contributes to understanding the complex nature of domain shifts in radar data and suggests paths forward for data collection strategy in the face of environmental variability.
- North America > United States (0.04)
- Europe > Slovenia > Drava > Municipality of Benedikt > Benedikt (0.04)
- Europe > Germany > Baden-Württemberg > Stuttgart Region > Stuttgart (0.04)
- Europe > Netherlands > South Holland > Delft (0.04)
Velocity Driven Vision: Asynchronous Sensor Fusion Birds Eye View Models for Autonomous Vehicles
Hayes, Seamie, Sharma, Sushil, Eising, Ciarán
Fusing different sensor modalities can be a difficult task, particularly if they are asynchronous. Asynchronisation may arise due to long processing times or improper synchronisation during calibration, and there must exist a way to still utilise this previous information for the purpose of safe driving, and object detection in ego vehicle/ multi-agent trajectory prediction. Difficulties arise in the fact that the sensor modalities have captured information at different times and also at different positions in space. Therefore, they are not spatially nor temporally aligned. This paper will investigate the challenge of radar and LiDAR sensors being asynchronous relative to the camera sensors, for various time latencies. The spatial alignment will be resolved before lifting into BEV space via the transformation of the radar/LiDAR point clouds into the new ego frame coordinate system. Only after this can we concatenate the radar/LiDAR point cloud and lifted camera features. Temporal alignment will be remedied for radar data only, we will implement a novel method of inferring the future radar point positions using the velocity information. Our approach to resolving the issue of sensor asynchrony yields promising results. We demonstrate velocity information can drastically improve IoU for asynchronous datasets, as for a time latency of 360 milliseconds (ms), IoU improves from 49.54 to 53.63. Additionally, for a time latency of 550ms, the camera+radar (C+R) model outperforms the camera+LiDAR (C+L) model by 0.18 IoU. This is an advancement in utilising the often-neglected radar sensor modality, which is less favoured than LiDAR for autonomous driving purposes.